fl method
- North America > United States > Virginia (0.04)
- North America > United States > Maryland (0.04)
Eliminating Domain Bias for Federated Learning in Representation Space
Recently, federated learning (FL) is popular for its privacy-preserving and collaborative learning abilities. However, under statistically heterogeneous scenarios, we observe that biased data domains on clients cause a representation bias phenomenon and further degenerate generic representations during local training, i.e., the representation degeneration phenomenon. To address these issues, we propose a general framework Domain Bias Eliminator (DBE) for FL. Our theoretical analysis reveals that DBE can promote bi-directional knowledge transfer between server and client, as it reduces the domain discrepancy between server and client in representation space. Besides, extensive experiments on four datasets show that DBE can greatly improve existing FL methods in both generalization and personalization abilities. The DBE-equipped FL method can outperform ten state-of-the-art personalized FL methods by a large margin. Our code is public at https://github.com/TsingZ0/DBE.
Federated Fine-tuning of Large Language Models under Heterogeneous Tasks and Client Resources
Federated Learning (FL) has recently been applied to the parameter-efficient fine-tuning of Large Language Models (LLMs). While promising, it raises significant challenges due to the heterogeneous resources and data distributions of clients.This study introduces FlexLoRA, a simple yet effective aggregation scheme for LLM fine-tuning, which mitigates the buckets effect in traditional FL that restricts the potential of clients with ample resources by tying them to the capabilities of the least-resourced participants. FlexLoRA allows for dynamic adjustment of local LoRA ranks, fostering the development of a global model imbued with broader, less task-specific knowledge.
- North America > United States > Virginia (0.04)
- Asia > China (0.04)
- South America > Peru > Lima Department > Lima Province > Lima (0.04)
- (2 more...)
- Information Technology > Security & Privacy (0.67)
- Education (0.66)
When to Stop Federated Learning: Zero-Shot Generation of Synthetic Validation Data with Generative AI for Early Stopping
Lee, Youngjoon, Lee, Hyukjoon, Gong, Jinu, Cao, Yang, Kang, Joonhyuk
Federated Learning (FL) enables collaborative model training across decentralized devices while preserving data privacy. However, FL methods typically run for a predefined number of global rounds, often leading to unnecessary computation when optimal performance is reached earlier. In addition, training may continue even when the model fails to achieve meaningful performance. To address this inefficiency, we introduce a zero-shot synthetic validation framework that leverages generative AI to monitor model performance and determine early stopping points. Our approach adaptively stops training near the optimal round, thereby conserving computational resources and enabling rapid hyperparameter adjustments. Numerical results on multi-label chest X-ray classification demonstrate that our method reduces training rounds by up to 74% while maintaining accuracy within 1% of the optimal.
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.14)
- Europe > Austria > Vienna (0.14)
- North America > United States > Maryland > Baltimore (0.04)
- (7 more...)
- Information Technology > Artificial Intelligence > Natural Language > Generation (0.73)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (0.71)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning > Generative AI (0.62)
Eliminating Domain Bias for Federated Learning in Representation Space
Recently, federated learning (FL) is popular for its privacy-preserving and collaborative learning abilities. However, under statistically heterogeneous scenarios, we observe that biased data domains on clients cause a representation bias phenomenon and further degenerate generic representations during local training, i.e ., the representation degeneration phenomenon.
- Asia > China > Shanghai > Shanghai (0.04)
- North America > United States > Virginia (0.04)
- North America > United States > Louisiana > East Baton Rouge Parish > Baton Rouge (0.04)
- (2 more...)
- Information Technology > Security & Privacy (1.00)
- Health & Medicine (0.93)
Supplementary Material
The supplementary material is organized as follows. First, we prove Proposition 1 and Theorem 1. In this section we prove Proposition 1, and some preliminary lemmas. Definition 4. Let the function Algorithm 1 for all i [m] and k 0. Let us define the following terms: g We will make use of the following notation for the history of the method. These samples are assumed to be independent across clients.
- North America > United States > Pennsylvania > Philadelphia County > Philadelphia (0.04)
- Asia > Middle East > Jordan (0.04)
GuardFed: A Trustworthy Federated Learning Framework Against Dual-Facet Attacks
Li, Yanli, Zhou, Yanan, Guo, Zhongliang, Yang, Nan, Zhang, Yuning, Chen, Huaming, Yuan, Dong, Ding, Weiping, Pedrycz, Witold
Abstract--Federated learning (FL) enables privacy-preserving collaborative model training but remains vulnerable to adversarial behaviors that compromise model utility or fairness across sensitive groups. While extensive studies have examined attacks targeting either objective, strategies that simultaneously degrade both utility and fairness remain largely unexplored. T o bridge this gap, we introduce the Dual-Facet Attack (DF A), a novel threat model that concurrently undermines predictive accuracy and group fairness. Two variants, Synchronous DF A (S-DF A) and Split DF A (Sp-DF A), are further proposed to capture distinct real-world collusion scenarios. Experimental results show that existing robust FL defenses, including hybrid aggregation schemes, fail to resist DF As effectively. T o counter these threats, we propose GuardFed, a self-adaptive defense framework that maintains a fairness-aware reference model using a small amount of clean server data augmented with synthetic samples. In each training round, GuardFed computes a dual-perspective trust score for every client by jointly evaluating its utility deviation and fairness degradation, thereby enabling selective aggregation of trustworthy updates. Extensive experiments on real-world datasets demonstrate that GuardFed consistently preserves both accuracy and fairness under diverse non-IID and adversarial conditions, achieving state-of-the-art performance compared with existing robust FL methods. The rapid advancement of deep learning (DL) has greatly accelerated the deployment of intelligent automation systems [1], providing smart services across diverse application domains. Alongside this evolution, there is an increasing emphasis on human-centered values such as privacy, fairness, and security, which extend beyond traditional performance-oriented objectives. Y anli Li is with the School of Artificial Intelligence and Computer Science, Nantong University, Nantong, 226019, China, and also with the School of Electrical and Computer Engineering, The University of Sydney, Sydney, 2006, Australia (e-mail: yanli.li@sydney.edu.au).
- Information Technology > Security & Privacy (1.00)
- Education (1.00)
High-Energy Concentration for Federated Learning in Frequency Domain
Shi, Haozhi, Xie, Weiying, Ye, Hangyu, Li, Daixun, Ma, Jitao, Li, Yunsong, Fang, Leyuan
Federated Learning (FL) presents significant potential for collaborative optimization without data sharing. Since synthetic data is sent to the server, leveraging the popular concept of dataset distillation, this FL framework protects real data privacy while alleviating data heterogeneity. However, such methods are still challenged by the redundant information and noise in entire spatial-domain designs, which inevitably increases the communication burden. In this paper, we propose a novel Frequency-Domain aware FL method with high-energy concentration (FedFD) to address this problem. Our FedFD is inspired by the discovery that the discrete cosine transform predominantly distributes energy to specific regions, referred to as high-energy concentration. The principle behind FedFD is that low-energy like high-frequency components usually contain redundant information and noise, thus filtering them helps reduce communication costs and optimize performance. Our FedFD is mathematically formulated to preserve the low-frequency components using a binary mask, facilitating an optimal solution through frequency-domain distribution alignment. In particular, real data-driven synthetic classification is imposed into the loss to enhance the quality of the low-frequency components. On five image and speech datasets, FedFD achieves superior performance than state-of-the-art methods while reducing communication costs. For example, on the CIFAR-10 dataset with Dirichlet coefficient $α= 0.01$, FedFD achieves a minimum reduction of 37.78\% in the communication cost, while attaining a 10.88\% performance gain.